Andaman and Nicobar Islands
New wolf snake honors the late Steve Irwin
Lycodon irwini is the latest species named after The Crocodile Hunter. Breakthroughs, discoveries, and DIY tips sent every weekday. Conservationists have discovered a previously unknown species of snake, slithering around one of Earth's most unique environments. In naming their new reptile, researchers decided to honor one of popular culture's most unique and beloved wildlife educators: the late, great Steve Irwin . The snake was discovered in the Nicobar Islands.
- Asia > India > Andaman and Nicobar Islands (0.27)
- Oceania > Australia > Queensland (0.05)
- North America > United States > New Jersey (0.05)
- (5 more...)
- Media (0.97)
- Health & Medicine > Therapeutic Area > Environmental Medicine > Snake Bites (0.71)
Integrating Linguistics and AI: Morphological Analysis and Corpus development of Endangered Toto Language of West Bengal
Guha, Ambalika, Saha, Sajal, Ballav, Debanjan, Mitra, Soumi, Chakraborty, Hritwick
Preserving linguistic diversity is necessary as every language offers a distinct perspective on the world. There have been numerous global initiatives to preserve endangered languages through documentation. This paper is a part of a project which aims to develop a trilingual (Toto-Bangla-English) language learning application to digitally archive and promote the endangered Toto language of West Bengal, India. This application, designed for both native Toto speakers and non-native learners, aims to revitalize the language by ensuring accessibility and usability through Unicode script integration and a structured language corpus. The research includes detailed linguistic documentation collected via fieldwork, followed by the creation of a morpheme-tagged, trilingual corpus used to train a Small Language Model (SLM) and a Transformer-based translation engine. The analysis covers inflectional morphology such as person-number-gender agreement, tense-aspect-mood distinctions, and case marking, alongside derivational strategies that reflect word-class changes. Script standardization and digital literacy tools were also developed to enhance script usage. The study offers a sustainable model for preserving endangered languages by incorporating traditional linguistic methodology with AI. This bridge between linguistic research with technological innovation highlights the value of interdisciplinary collaboration for community-based language revitalization.
- Asia > India > West Bengal > Kolkata (0.14)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.05)
- Asia > China (0.04)
- (14 more...)
- Media > Film (1.00)
- Leisure & Entertainment (1.00)
- Education (1.00)
Reinforcing Multi-Turn Reasoning in LLM Agents via Turn-Level Reward Design
Wei, Quan, Zeng, Siliang, Li, Chenliang, Brown, William, Frunza, Oana, Deng, Wei, Schneider, Anderson, Nevmyvaka, Yuriy, Zhao, Yang Katie, Garcia, Alfredo, Hong, Mingyi
This paper investigates Reinforcement Learning (RL) approaches to enhance the reasoning capabilities of Large Language Model (LLM) agents in long-horizon, multi-turn scenarios. Although RL algorithms such as Group Relative Policy Optimization (GRPO) and Proximal Policy Optimization (PPO) have been widely applied to train multi-turn LLM agents, they typically rely only on sparse outcome rewards and lack dense intermediate signals across multiple decision steps, limiting their performance on complex reasoning tasks. To bridge this gap, we present the first systematic study of \textit{turn-level reward design} for multi-turn RL algorithms and agent applications. By integrating turn-level rewards, we extend GRPO and PPO to their respective multi-turn variants, enabling fine-grained credit assignment. We conduct case studies on multi-turn reasoning-augmented search agents, where we carefully design two types of turn-level rewards: verifiable and LLM-as-judge. Our experiments on multi-turn search tasks demonstrate that incorporating well-designed turn-level rewards enables RL algorithms to significantly outperform baseline methods with trajectory-level rewards. Both training and validation reward curves illustrate that our method achieves \textit{greater stability}, \textit{faster convergence}, and \textit{higher accuracy}. Numerical results across diverse question-answering datasets further show that our approach consistently delivers highest answer correctness and 100\% format correctness.
- Europe > United Kingdom (0.28)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > Texas (0.04)
- (6 more...)
Evaluating Large Language Models for IUCN Red List Species Information
Large Language Models (LLMs) are rapidly being adopted in conservation to address the biodiversity crisis, yet their reliability for species evaluation is uncertain. This study systematically validates five leading models on 21,955 species across four core IUCN Red List assessment components: taxonomy, conservation status, distribution, and threats. A critical paradox was revealed: models excelled at taxonomic classification (94.9%) but consistently failed at conservation reasoning (27.2% for status assessment). This knowledge-reasoning gap, evident across all models, suggests inherent architectural constraints, not just data limitations. Furthermore, models exhibited systematic biases favoring charismatic vertebrates, potentially amplifying existing conservation inequities. These findings delineate clear boundaries for responsible LLM deployment: they are powerful tools for information retrieval but require human oversight for judgment-based decisions. A hybrid approach is recommended, where LLMs augment expert capacity while human experts retain sole authority over risk assessment and policy.
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.70)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.93)
Better To Ask in English? Evaluating Factual Accuracy of Multilingual LLMs in English and Low-Resource Languages
Rohera, Pritika, Ginimav, Chaitrali, Sawant, Gayatri, Joshi, Raviraj
Multilingual Large Language Models (LLMs) have demonstrated significant effectiveness across various languages, particularly in high-resource languages such as English. However, their performance in terms of factual accuracy across other low-resource languages, especially Indic languages, remains an area of investigation. In this study, we assess the factual accuracy of LLMs - GPT-4o, Gemma-2-9B, Gemma-2-2B, and Llama-3.1-8B - by comparing their performance in English and Indic languages using the IndicQuest dataset, which contains question-answer pairs in English and 19 Indic languages. By asking the same questions in English and their respective Indic translations, we analyze whether the models are more reliable for regional context questions in Indic languages or when operating in English. Our findings reveal that LLMs often perform better in English, even for questions rooted in Indic contexts. Notably, we observe a higher tendency for hallucination in responses generated in low-resource Indic languages, highlighting challenges in the multilingual understanding capabilities of current LLMs.
- Asia > Thailand > Bangkok > Bangkok (0.04)
- Asia > India > West Bengal (0.04)
- Asia > India > Uttarakhand (0.04)
- (21 more...)
Millions of $\text{GeAR}$-s: Extending GraphRAG to Millions of Documents
Shen, Zhili, Diao, Chenxin, Merita, Pascual, Vougiouklis, Pavlos, Pan, Jeff Z.
Recent studies have explored graph-based approaches to retrieval-augmented generation, leveraging structured or semi-structured information -- such as entities and their relations extracted from documents -- to enhance retrieval. However, these methods are typically designed to address specific tasks, such as multi-hop question answering and query-focused summarisation, and therefore, there is limited evidence of their general applicability across broader datasets. In this paper, we aim to adapt a state-of-the-art graph-based RAG solution: $\text{GeAR}$ and explore its performance and limitations on the SIGIR 2025 LiveRAG Challenge.
- Asia > India > Andaman and Nicobar Islands (0.14)
- Europe > Italy (0.05)
- Europe > United Kingdom > Scotland > City of Edinburgh > Edinburgh (0.05)
- (11 more...)
- Leisure & Entertainment (1.00)
- Health & Medicine (1.00)
- Media > Film (0.96)
- Education (0.94)
Improving Bangla Linguistics: Advanced LSTM, Bi-LSTM, and Seq2Seq Models for Translating Sylheti to Modern Bangla
Das, Sourav Kumar, Naeen, Md. Julkar, Islam, MD. Jahidul, Sajeeb, Md. Anisul Haque, Chakraborty, Narayan Ranjan, Mojumdar, Mayen Uddin
Bangla or Bengali is the national language of Bangladesh, people from different regions don't talk in proper Bangla. Every division of Bangladesh has its own local language like Sylheti, Chittagong etc. In recent years some papers were published on Bangla language like sentiment analysis, fake news detection and classifications, but a few of them were on Bangla languages. This research is for the local language and this particular paper is on Sylheti language. It presented a comprehensive system using Natural Language Processing or NLP techniques for translating Pure or Modern Bangla to locally spoken Sylheti Bangla language. Total 1200 data used for training 3 models LSTM, Bi-LSTM and Seq2Seq and LSTM scored the best in performance with 89.3% accuracy. The findings of this research may contribute to the growth of Bangla NLP researchers for future more advanced innovations.
- Asia > Bangladesh > Dhaka Division > Dhaka District > Dhaka (0.05)
- Asia > Singapore (0.04)
- Asia > India > West Bengal (0.04)
- (2 more...)
Data Driven Deep Learning for Correcting Global Climate Model Projections of SST and DSL in the Bay of Bengal
Pasula, Abhishek, Subramani, Deepak N.
Climate change alters ocean conditions, notably temperature and sea level. In the Bay of Bengal, these changes influence monsoon precipitation and marine productivity, critical to the Indian economy. In Phase 6 of the Coupled Model Intercomparison Project (CMIP6), Global Climate Models (GCMs) use different shared socioeconomic pathways (SSPs) to obtain future climate projections. However, significant discrepancies are observed between these models and the reanalysis data in the Bay of Bengal for 2015-2024. Specifically, the root mean square error (RMSE) between the climate model output and the Ocean Reanalysis System (ORAS5) is 1.2C for the sea surface temperature (SST) and 1.1 m for the dynamic sea level (DSL). We introduce a new data-driven deep learning model to correct for this bias. The deep neural model for each variable is trained using pairs of climatology-removed monthly climate projections as input and the corresponding month's ORAS5 as output. This model is trained with historical data (1950 to 2014), validated with future projection data from 2015 to 2020, and tested with future projections from 2021 to 2023. Compared to the conventional EquiDistant Cumulative Distribution Function (EDCDF) statistical method for bias correction in climate models, our approach decreases RMSE by 0.15C for SST and 0.3 m for DSL. The trained model subsequently corrects the projections for 2024-2100. A detailed analysis of the monthly, seasonal, and decadal means and variability is performed to underscore the implications of the novel dynamics uncovered in our corrected projections.
- Asia > Sri Lanka (0.04)
- Asia > India > Andaman and Nicobar Islands (0.04)
- North America > Saint Martin (0.04)
- (8 more...)
Agricultural Landscape Understanding At Country-Scale
Dua, Radhika, Saxena, Nikita, Agarwal, Aditi, Wilson, Alex, Singh, Gaurav, Tran, Hoang, Deshpande, Ishan, Kaur, Amandeep, Aggarwal, Gaurav, Nath, Chandan, Basu, Arnab, Batchu, Vishal, Holla, Sharath, Kurle, Bindiya, Missura, Olana, Aggarwal, Rahul, Garg, Shubhika, Shah, Nishi, Singh, Avneet, Tewari, Dinesh, Dondzik, Agata, Adsul, Bharat, Sohoni, Milind, Praveen, Asim Rama, Dangi, Aaryan, Kadivar, Lisan, Abhishek, E, Sudhansu, Niranjan, Hattekar, Kamlakar, Datar, Sameer, Chaithanya, Musty Krishna, Reddy, Anumas Ranjith, Kumar, Aashish, Tirumala, Betala Laxmi, Talekar, Alok
The global food system is facing unprecedented challenges. In 2023, 2.4 billion people experienced moderate to severe food insecurity [1], a crisis precipitated by anthropogenic climate change and evolving dietary preferences. Furthermore, the food system itself significantly contributes to the climate crisis, with food loss and waste accounting for 2.4 gigatonnes of carbon dioxide equivalent emissions per year (GT CO2e/yr) [2], and the production, mismanagement, and misapplication of agricultural inputs such as fertilizers and manure generating an additional 2.5 GT CO2e/yr [3]. To sustain a projected global population of 9.6 billion by 2050, the Food and Agriculture Organization (FAO) estimates that food production must increase by at least 60% [1]. However, this also presents an opportunity: transitioning to sustainable agricultural practices can transform the sector from a net source of greenhouse gas emissions to a vital carbon sink.
- Asia > India > Andaman and Nicobar Islands (0.14)
- Asia > India > Telangana (0.05)
- Asia > India > Maharashtra (0.05)
- (25 more...)
- Food & Agriculture > Agriculture (1.00)
- Materials > Chemicals > Agricultural Chemicals (0.34)
Think-on-Graph 2.0: Deep and Interpretable Large Language Model Reasoning with Knowledge Graph-guided Retrieval
Ma, Shengjie, Xu, Chengjin, Jiang, Xuhui, Li, Muzhi, Qu, Huaren, Guo, Jian
Retrieval-augmented generation (RAG) has significantly advanced large language models (LLMs) by enabling dynamic information retrieval to mitigate knowledge gaps and hallucinations in generated content. However, these systems often falter with complex reasoning and consistency across diverse queries. In this work, we present Think-on-Graph 2.0, an enhanced RAG framework that aligns questions with the knowledge graph and uses it as a navigational tool, which deepens and refines the RAG paradigm for information collection and integration. The KG-guided navigation fosters deep and long-range associations to uphold logical consistency and optimize the scope of retrieval for precision and interoperability. In conjunction, factual consistency can be better ensured through semantic similarity guided by precise directives. ToG${2.0}$ not only improves the accuracy and reliability of LLMs' responses but also demonstrates the potential of hybrid structured knowledge systems to significantly advance LLM reasoning, aligning it closer to human-like performance. We conducted extensive experiments on four public datasets to demonstrate the advantages of our method compared to the baseline.
- Asia > India > Andaman and Nicobar Islands (0.14)
- North America > United States > District of Columbia > Washington (0.05)
- Indian Ocean > Bay of Bengal (0.05)
- (9 more...)